skip to main content


Search for: All records

Creators/Authors contains: "Narayanan, Shrikanth"

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Free, publicly-accessible full text available June 13, 2024
  2. Free, publicly-accessible full text available June 8, 2024
  3. Individuals who have undergone treatment for oral cancer oftentimes exhibit compensatory behavior in consonant production. This pilot study investigates whether compensatory mechanisms utilized in the production of speech sounds with a given target constriction location vary systematically depending on target manner of articulation. The data reveal that compensatory strategies used to produce target alveolar segments vary systematically as a function of target manner of articulation in subtle yet meaningful ways. When target constriction degree at a particular constriction location cannot be preserved, individuals may leverage their ability to finely modulate constriction degree at multiple constriction locations along the vocal tract. 
    more » « less
  4. The theory of Task Dynamics provides a method of predicting articulatory kinematics from a discrete phonologically-relevant representation (“gestural score”). However, because the implementations of that model (e.g., Nam et al., 2004) have generally used a simplified articulatory geometry (Mermelstein et al., 1981) whose forward model (from articulator to constriction coordinates) can be analytically derived, quantitative predictions of the model for individual human vocal tracts have not been possible. Recently, methods of deriving individual speaker forward models from real-time MRI data have been developed (Sorensen et al., 2019). This has further allowed development of task dynamic models for individual speakers, which make quantitative predictions. Thus far, however, these models (Alexander et al., 2019) could only synthesize limited types of utterances due to their inability to model temporally overlapping gestures. An updated implementation is presented, which can accommodate overlapping gestures and incorporates an optimization loop to improve the fit of modeled articulatory trajectories to the observed ones. Using an analysis-by-synthesis approach, the updated implementation can be utilized: (1) to refine the hypothesized speaker-general gestural parameters (target, stiffness) for individual speakers; (2) to test different degrees of temporal overlapping among multiple gestures such as a CCVC syllable. [Work supported by NSF, Grant 1908865.]

     
    more » « less
  5. Lighting, as a significant component of indoor environment quality, was found to be a primary contributor to deficient indoor environments in today’s workplace. This resulted from the fact that current guidelines are derived from empirical values and neglect the prevalence of computer-based tasks in current offices. A personal visual comfort model was designed to predict the degree of an individual’s visual comfort, as a way of evaluating the indoor lighting of the environment. Development of the model relied on experimental data, including individual eye pupil sizes, visual sensations, and visual satisfaction in response to various illuminance levels used for tests of six human subjects. The results showed that (1) A personal comfort model was needed, (2) the personal comfort model produced a median accuracy of 0.7086 for visual sensation and 0.65467 for visual satisfaction for all subjects; (3) To develop a prediction model for the sample group, the Support Vector Machine algorithm,, outperformed the Logistic Regression and the Gaussian Naïve Bayes in terms of prediction accuracy. It was concluded that, a personal visual comfort model can use a building’s occupant’s eye pupil size to generate an accurate prediction of that occupant’s visual sensations and visual satisfaction that can, therefore, be applied with lighting control to improve the indoor environment and energy use in that building. 
    more » « less
  6. Abstract

    Real-time magnetic resonance imaging (RT-MRI) of human speech production is enabling significant advances in speech science, linguistics, bio-inspired speech technology development, and clinical applications. Easy access to RT-MRI is however limited, and comprehensive datasets with broad access are needed to catalyze research across numerous domains. The imaging of the rapidly moving articulators and dynamic airway shaping during speech demands high spatio-temporal resolution and robust reconstruction methods. Further, while reconstructed images have been published, to-date there is no open dataset providing raw multi-coil RT-MRI data from an optimized speech production experimental setup. Such datasets could enable new and improved methods for dynamic image reconstruction, artifact correction, feature extraction, and direct extraction of linguistically-relevant biomarkers. The present dataset offers a unique corpus of 2D sagittal-view RT-MRI videos along with synchronized audio for 75 participants performing linguistically motivated speech tasks, alongside the corresponding public domain raw RT-MRI data. The dataset also includes 3D volumetric vocal tract MRI during sustained speech sounds and high-resolution static anatomical T2-weighted upper airway MRI for each participant.

     
    more » « less